A Correlation Analysis on LSA and HAL Semantic Space Models

نویسندگان

  • Xin Yan
  • Xue Li
  • Dawei Song
چکیده

In this paper, we compare a well-known semantic space model, Latent Semantic Analysis (LSA) with another model, Hyperspace Analogue to Language (HAL) which is widely used in different area, especially in automatic query refinement. We conduct this comparative analysis to prove our hypothesis that with respect to ability of extracting the lexical information from a corpus of text, LSA is quite similar to HAL. We regard HAL and LSA as black boxes. Through a Pearson’s correlation analysis to the outputs of these two black boxes, we conclude that LSA highly co-relates with HAL and thus there is a justification that LSA and HAL can potentially play a similar role in the area of facilitating automatic query refinement. This paper evaluates LSA in a new application area and contributes an effective way to compare different semantic space models.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Predicting Lexical Priming Effects from Distributional Semantic Similarities: A Replication with Extension

In two experiments, we attempted to replicate and extend findings by Günther et al. (2016) that word similarity measures obtained from distributional semantics models-Latent Semantic Analysis (LSA) and Hyperspace Analog to Language (HAL)-predict lexical priming effects. To this end, we used the pseudo-random method to generate item material while systematically controlling for word similarities...

متن کامل

Computing Semantic Representations: A Comparative Analysis

How can we formally capture the complex semantic relationships of the human lexicon? This question has been the focus of much recent computational studies. The ability to represent semantics faithfully in formal mechanisms not only is important for understanding the nature of the lexical system of natural languages, but also has significant implications for understanding the mental representati...

متن کامل

Latent Semantic Analysis based Language Models for Meetings

Language models that combine N -gram models with Latent Semantic Analysis (LSA) based models have been successfully applied for conversational speech recognition [3] and for the Wall Street Journal recognition task [1]. LSA defines a semantic similarity space using a training corpus. This semantic similarity can be used for dealing with long distance dependencies, which are an inherent problem ...

متن کامل

On Automatic Annotation of Images with Latent Space Models

Image auto-annotation, i.e., the association of words to whole images, has attracted considerable attention. In particular, unsupervised, probabilistic latent variable models of text and image features have shown encouraging results, but their performance with respect to other approaches remains unknown. In this paper, we apply and compare two simple latent space models commonly used in text an...

متن کامل

Combination of latent semantic analysis based language models for meeting recognition

Latent Semantic Analysis (LSA) defines a semantic similarity space using a training corpus. This semantic similarity can be used for dealing with long distance dependencies, which are an inherent problem for traditional wordbased n-gram models. Since LSA models adapt dynamically to topics, and meetings have clear topics, we conjecture that these models can improve speech recognition accuracy on...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004